2024-05-08 08:33:03
This https://arxiv.org/abs/2405.02937 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2405.02937 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
Understanding Language Modeling Paradigm Adaptations in Recommender Systems: Lessons Learned and Open Challenges
Lemei Zhang, Peng Liu, Yashar Deldjoo, Yong Zheng, Jon Atle Gulla
https://arxiv.org/abs/2404.03788
Improving Long Text Understanding with Knowledge Distilled from Summarization Model
Yan Liu, Yazheng Yang, Xiaokang Chen
https://arxiv.org/abs/2405.04955 h…
This https://arxiv.org/abs/2310.12357 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csSE_…
Improving Long Text Understanding with Knowledge Distilled from Summarization Model
Yan Liu, Yazheng Yang, Xiaokang Chen
https://arxiv.org/abs/2405.04955 h…
This https://arxiv.org/abs/2404.13236 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csDC_…
TTPXHunter: Actionable Threat Intelligence Extraction as TTPs form Finished Cyber Threat Reports
Nanda Rani, Bikash Saha, Vikas Maurya, Sandeep Kumar Shukla
https://arxiv.org/abs/2403.03267
ProLLaMA: A Protein Large Language Model for Multi-Task Protein Language Processing
Liuzhenghao Lv, Zongying Lin, Hao Li, Yuyang Liu, Jiaxi Cui, Calvin Yu-Chian Chen, Li Yuan, Yonghong Tian
https://arxiv.org/abs/2402.16445 https://arxiv.org/pdf/2402.16445
arXiv:2402.16445v1 Announce Type: new
Abstract: Large Language Models (LLMs), including GPT-x and LLaMA2, have achieved remarkable performance in multiple Natural Language Processing (NLP) tasks. Under the premise that protein sequences constitute the protein language, Protein Large Language Models (ProLLMs) trained on protein corpora excel at de novo protein sequence generation. However, as of now, unlike LLMs in NLP, no ProLLM is capable of multiple tasks in the Protein Language Processing (PLP) field. This prompts us to delineate the inherent limitations in current ProLLMs: (i) the lack of natural language capabilities, (ii) insufficient instruction understanding, and (iii) high training resource demands. To address these challenges, we introduce a training framework to transform any general LLM into a ProLLM capable of handling multiple PLP tasks. Specifically, our framework utilizes low-rank adaptation and employs a two-stage training approach, and it is distinguished by its universality, low overhead, and scalability. Through training under this framework, we propose the ProLLaMA model, the first known ProLLM to handle multiple PLP tasks simultaneously. Experiments show that ProLLaMA achieves state-of-the-art results in the unconditional protein sequence generation task. In the controllable protein sequence generation task, ProLLaMA can design novel proteins with desired functionalities. In the protein property prediction task, ProLLaMA achieves nearly 100\% accuracy across many categories. The latter two tasks are beyond the reach of other ProLLMs. Code is available at \url{https://github.com/Lyu6PosHao/ProLLaMA}.
This https://arxiv.org/abs/2403.01528 has been replaced.
link: https://scholar.google.com/scholar?q=a
Do Large Language Models Rank Fairly? An Empirical Study on the Fairness of LLMs as Rankers
Yuan Wang, Xuyang Wu, Hsin-Tai Wu, Zhiqiang Tao, Yi Fang
https://arxiv.org/abs/2404.03192
This https://arxiv.org/abs/2309.12284 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2306.11943 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csSE_…
This https://arxiv.org/abs/2308.06013 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csIT_…
Vision-Language Navigation with Embodied Intelligence: A Survey
Peng Gao, Peng Wang, Feng Gao, Fei Wang, Ruyue Yuan
https://arxiv.org/abs/2402.14304 https:…
Application of GPT Language Models for Innovation in Activities in University Teaching
Manuel de Buenaga, Francisco Javier Bueno
https://arxiv.org/abs/2403.14694
Insights into Natural Language Database Query Errors: From Attention Misalignment to User Handling Strategies
Zheng Ning, Yuan Tian, Zheng Zhang, Tianyi Zhang, Toby Li
https://arxiv.org/abs/2402.07304
This https://arxiv.org/abs/2310.06555 has been replaced.
link: https://scholar.google.com/scholar?q=a
LLMChain: Blockchain-based Reputation System for Sharing and Evaluating Large Language Models
Mouhamed Amine Bouchiha, Quentin Telnoff, Souhail Bakkali, Ronan Champagnat, Mourad Rabah, Micka\"el Coustaty, Yacine Ghamri-Doudane
https://arxiv.org/abs/2404.13236
Vision-Language Navigation with Embodied Intelligence: A Survey
Peng Gao, Peng Wang, Feng Gao, Fei Wang, Ruyue Yuan
https://arxiv.org/abs/2402.14304 https:…
Pegasus-v1 Technical Report
Raehyuk Jung, Hyojun Go, Jaehyuk Yi, Jiho Jang, Daniel Kim, Jay Suh, Aiden Lee, Cooper Han, Jae Lee, Jeff Kim, Jin-Young Kim, Junwan Kim, Kyle Park, Lucas Lee, Mars Ha, Minjoon Seo, Abraham Jo, Ed Park, Hassan Kianinejad, SJ Kim, Tony Moon, Wade Jeong, Andrei Popescu, Esther Kim, EK Yoon, Genie Heo, Henry Choi, Jenna Kang, Kevin Han, Noah Seo, Sunny Nguyen, Ryan Won, Yeonhoo Park, Anthony Giuliani, Dave Chung, Hans Yoon, James Le, Jenny Ahn, June Lee, Manind…
This https://arxiv.org/abs/2403.04786 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCR_…
From Perils to Possibilities: Understanding how Human (and AI) Biases affect Online Fora
Virginia Morini, Valentina Pansanella, Katherine Abramski, Erica Cau, Andrea Failla, Salvatore Citraro, Giulio Rossetti
https://arxiv.org/abs/2403.14298
MineDreamer: Learning to Follow Instructions via Chain-of-Imagination for Simulated-World Control
Enshen Zhou, Yiran Qin, Zhenfei Yin, Yuzhou Huang, Ruimao Zhang, Lu Sheng, Yu Qiao, Jing Shao
https://arxiv.org/abs/2403.12037
Training Table Question Answering via SQL Query Decomposition
Rapha\"el Mouravieff, Benjamin Piwowarski, Sylvain Lamprier
https://arxiv.org/abs/2402.13288
Saving the legacy of Hero Ibash: Evaluating Four Language Models for Aminoacian
Yunze Xiao, Yiyang Pan
https://arxiv.org/abs/2402.18121 https://
RAG and RAU: A Survey on Retrieval-Augmented Language Model in Natural Language Processing
Yucheng Hu, Yuxing Lu
https://arxiv.org/abs/2404.19543 https://arxiv.org/pdf/2404.19543
arXiv:2404.19543v1 Announce Type: new
Abstract: Large Language Models (LLMs) have catalyzed significant advancements in Natural Language Processing (NLP), yet they encounter challenges such as hallucination and the need for domain-specific knowledge. To mitigate these, recent methodologies have integrated information retrieved from external resources with LLMs, substantially enhancing their performance across NLP tasks. This survey paper addresses the absence of a comprehensive overview on Retrieval-Augmented Language Models (RALMs), both Retrieval-Augmented Generation (RAG) and Retrieval-Augmented Understanding (RAU), providing an in-depth examination of their paradigm, evolution, taxonomy, and applications. The paper discusses the essential components of RALMs, including Retrievers, Language Models, and Augmentations, and how their interactions lead to diverse model structures and applications. RALMs demonstrate utility in a spectrum of tasks, from translation and dialogue systems to knowledge-intensive applications. The survey includes several evaluation methods of RALMs, emphasizing the importance of robustness, accuracy, and relevance in their assessment. It also acknowledges the limitations of RALMs, particularly in retrieval quality and computational efficiency, offering directions for future research. In conclusion, this survey aims to offer a structured insight into RALMs, their potential, and the avenues for their future development in NLP. The paper is supplemented with a Github Repository containing the surveyed works and resources for further study: https://github.com/2471023025/RALM_Survey.
Multimodal Human-Autonomous Agents Interaction Using Pre-Trained Language and Visual Foundation Models
Linus Nwankwo, Elmar Rueckert
https://arxiv.org/abs/2403.12273
Efficient and Scalable Fine-Tune of Language Models for Genome Understanding
Huixin Zhan, Ying Nian Wu, Zijun Zhang
https://arxiv.org/abs/2402.08075 https:…
Securing Large Language Models: Threats, Vulnerabilities and Responsible Practices
Sara Abdali, Richard Anarfi, CJ Barberan, Jia He
https://arxiv.org/abs/2403.12503
Perplexed: Understanding When Large Language Models are Confused
Nathan Cooper, Torsten Scholak
https://arxiv.org/abs/2404.06634 https://
This https://arxiv.org/abs/2402.18838 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
Measuring Geographic Diversity of Foundation Models with a Natural Language--based Geo-guessing Experiment on GPT-4
Zilong Liu, Krzysztof Janowicz, Kitty Currier, Meilin Shi
https://arxiv.org/abs/2404.07612
This https://arxiv.org/abs/2402.18838 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2312.02003 has been replaced.
link: https://scholar.google.com/scholar?q=a
Enabling Waypoint Generation for Collaborative Robots using LLMs and Mixed Reality
Cathy Mengying Fang, Krzysztof Zieli\'nski, Pattie Maes, Joe Paradiso, Bruce Blumberg, Mikkel Baun Kj{\ae}rgaard
https://arxiv.org/abs/2403.09308
Reconfigurable Robot Identification from Motion Data
Yuhang Hu, Yunzhe Wang, Ruibo Liu, Zhou Shen, Hod Lipson
https://arxiv.org/abs/2403.10496 https://
The First Place Solution of WSDM Cup 2024: Leveraging Large Language Models for Conversational Multi-Doc QA
Yiming Li, Zhao Zhang
https://arxiv.org/abs/2402.18385
This https://arxiv.org/abs/2307.08309 has been replaced.
link: https://scholar.google.com/scholar?q=a
RITFIS: Robust input testing framework for LLMs-based intelligent software
Mingxuan Xiao, Yan Xiao, Hai Dong, Shunhui Ji, Pengcheng Zhang
https://arxiv.org/abs/2402.13518
This https://arxiv.org/abs/2209.01678 has been replaced.
link: https://scholar.google.com/scholar?q=a
This https://arxiv.org/abs/2402.14304 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csRO_…
Incorporating Graph Attention Mechanism into Geometric Problem Solving Based on Deep Reinforcement Learning
Xiuqin Zhong, Shengyuan Yan, Gongqi Lin, Hongguang Fu, Liang Xu, Siwen Jiang, Lei Huang, Wei Fang
https://arxiv.org/abs/2403.14690
This https://arxiv.org/abs/2402.08015 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2402.11194 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
CONLINE: Complex Code Generation and Refinement with Online Searching and Correctness Testing
Xinyi He, Jiaru Zou, Yun Lin, Mengyu Zhou, Shi Han, Zejian Yuan, Dongmei Zhang
https://arxiv.org/abs/2403.13583
Towards Better Understanding of Contrastive Sentence Representation Learning: A Unified Paradigm for Gradient
Mingxin Li, Richong Zhang, Zhijie Nie
https://arxiv.org/abs/2402.18281
CONLINE: Complex Code Generation and Refinement with Online Searching and Correctness Testing
Xinyi He, Jiaru Zou, Yun Lin, Mengyu Zhou, Shi Han, Zejian Yuan, Dongmei Zhang
https://arxiv.org/abs/2403.13583
This https://arxiv.org/abs/2401.01085 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCR_…
Beyond the Headlines: Understanding Sentiments and Morals Impacting Female Employment in Spain
Oscar Araque, Luca Barbaglia, Francesco Berlingieri, Marco Colagrossi, Sergio Consoli, Lorenzo Gatti, Caterina Mauri, Kyriaki Kalimeri
https://arxiv.org/abs/2402.07339
Quantifying Contamination in Evaluating Code Generation Capabilities of Language Models
Martin Riddell, Ansong Ni, Arman Cohan
https://arxiv.org/abs/2403.04811
Breaking Down the Defenses: A Comparative Survey of Attacks on Large Language Models
Arijit Ghosh Chowdhury, Md Mofijul Islam, Vaibhav Kumar, Faysal Hossain Shezan, Vaibhav Kumar, Vinija Jain, Aman Chadha
https://arxiv.org/abs/2403.04786
From Pixels to Insights: A Survey on Automatic Chart Understanding in the Era of Large Foundation Models
Kung-Hsiang Huang, Hou Pong Chan, Yi R. Fung, Haoyi Qiu, Mingyang Zhou, Shafiq Joty, Shih-Fu Chang, Heng Ji
https://arxiv.org/abs/2403.12027
Breaking Down the Defenses: A Comparative Survey of Attacks on Large Language Models
Arijit Ghosh Chowdhury, Md Mofijul Islam, Vaibhav Kumar, Faysal Hossain Shezan, Vaibhav Kumar, Vinija Jain, Aman Chadha
https://arxiv.org/abs/2403.04786
This https://arxiv.org/abs/2402.05130 has been replaced.
link: https://scholar.google.com/scholar?q=a
This https://arxiv.org/abs/2402.05130 has been replaced.
link: https://scholar.google.com/scholar?q=a
This https://arxiv.org/abs/2403.03267 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCR_…
This https://arxiv.org/abs/2311.13668 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
Large Language Models: A Survey
Shervin Minaee, Tomas Mikolov, Narjes Nikzad, Meysam Chenaghlu, Richard Socher, Xavier Amatriain, Jianfeng Gao
https://arxiv.org/abs/2402.06196
This https://arxiv.org/abs/2401.09615 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
NLP4RE Tools: Classification, Overview, and Management
Julian Frattini, Michael Unterkalmsteiner, Davide Fucci, Daniel Mendez
https://arxiv.org/abs/2403.06685
This https://arxiv.org/abs/2403.03267 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCR_…
This https://arxiv.org/abs/2402.12243 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2312.15918 has been replaced.
link: https://scholar.google.com/scholar?q=a
This https://arxiv.org/abs/2310.14174 has been replaced.
link: https://scholar.google.com/scholar?q=a
This https://arxiv.org/abs/2309.08968 has been replaced.
link: https://scholar.google.com/scholar?q=a
This https://arxiv.org/abs/2309.08968 has been replaced.
link: https://scholar.google.com/scholar?q=a
Cross-lingual Transfer or Machine Translation? On Data Augmentation for Monolingual Semantic Textual Similarity
Sho Hoshino, Akihiko Kato, Soichiro Murakami, Peinan Zhang
https://arxiv.org/abs/2403.05257
Cross-lingual Transfer or Machine Translation? On Data Augmentation for Monolingual Semantic Textual Similarity
Sho Hoshino, Akihiko Kato, Soichiro Murakami, Peinan Zhang
https://arxiv.org/abs/2403.05257
This https://arxiv.org/abs/2309.08345 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2309.08345 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
LLMs' Reading Comprehension Is Affected by Parametric Knowledge and Struggles with Hypothetical Statements
Victoria Basmov, Yoav Goldberg, Reut Tsarfaty
https://arxiv.org/abs/2404.06283
This https://arxiv.org/abs/2402.12730 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
FaBERT: Pre-training BERT on Persian Blogs
Mostafa Masumi, Seyed Soroush Majd, Mehrnoush Shamsfard, Hamid Beigy
https://arxiv.org/abs/2402.06617 https://…
ERA-CoT: Improving Chain-of-Thought through Entity Relationship Analysis
Yanming Liu, Xinyue Peng, Tianyu Du, Jianwei Yin, Weihao Liu, Xuhong Zhang
https://arxiv.org/abs/2403.06932
Finding fake reviews in e-commerce platforms by using hybrid algorithms
Mathivanan Periasamy, Rohith Mahadevan, Bagiya Lakshmi S, Raja CSP Raman, Hasan Kumar S, Jasper Jessiman
https://arxiv.org/abs/2404.06339
This https://arxiv.org/abs/2311.13668 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2311.13668 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…